Goto

Collaborating Authors

 Lucknow


Beyond the Hype: Comparing Lightweight and Deep Learning Models for Air Quality Forecasting

Gondal, Moazzam Umer, Qudous, Hamad ul, Farhan, Asma Ahmad

arXiv.org Machine Learning

Accurate forecasting of urban air pollution is essential for protecting public health and guiding mitigation policies. While Deep Learning (DL) and hybrid pipelines dominate recent research, their complexity and limited interpretability hinder operational use. This study investigates whether lightweight additive models -- Facebook Prophet (FBP) and NeuralProphet (NP) -- can deliver competitive forecasts for particulate matter (PM$_{2.5}$, PM$_{10}$) in Beijing, China. Using multi-year pollutant and meteorological data, we applied systematic feature selection (correlation, mutual information, mRMR), leakage-safe scaling, and chronological data splits. Both models were trained with pollutant and precursor regressors, with NP additionally leveraging lagged dependencies. For context, two machine learning baselines (LSTM, LightGBM) and one traditional statistical model (SARIMAX) were also implemented. Performance was evaluated on a 7-day holdout using MAE, RMSE, and $R^2$. Results show that FBP consistently outperformed NP, SARIMAX, and the learning-based baselines, achieving test $R^2$ above 0.94 for both pollutants. These findings demonstrate that interpretable additive models remain competitive with both traditional and complex approaches, offering a practical balance of accuracy, transparency, and ease of deployment.


Provably Outlier-resistant Semi-parametric Regression for Transferable Calibration of Low-cost Air-quality Sensors

Chaurasia, Divyansh, Daram, Manoj, Kumar, Roshan, Rao, Nihal Thukarama, Sangode, Vipul, Srivastava, Pranjal, Tripathi, Avnish, Chakraborty, Shoubhik, Akanksha, null, Kumar, Ambasht, Sethi, Davender, Tripathi, Sachchida Nand, Kar, Purushottam

arXiv.org Machine Learning

LCAQ sensors have been shown to play a critical role in the establishment of dense, expansive air-quality monitoring networks and combating elevated pollution levels. The calibration of LCAQ sensors against regulatory-grade monitors is an expensive, laborious and time-consuming process, especially when a large number of sensors are to be deployed in a geographically diverse layout. In this work, we present the RESPIRE technique to calibrate LCAQ sensors to detect ambient CO (Carbon Monoxide) levels. RESPIRE offers specific advantages over baseline calibration methods popular in literature, such as improved prediction in cross-site, cross-season, and cross-sensor settings. RESPIRE offers a training algorithm that is provably resistant to outliers and an explainable model with the ability to flag instances of model overfitting. Empirical results are presented based on data collected during an extensive deployment spanning four sites, two seasons and six sensor packages.


Inclusion of Role into Named Entity Recognition and Ranking

Shukla, Neelesh Kumar, Singh, Sanasam Ranbir

arXiv.org Artificial Intelligence

Most of the Natural Language Processing systems are involved in entity-based processing for several tasks like Information Extraction, Question-Answering, Text-Summarization and so on. A new challenge comes when entities play roles according to their act or attributes in certain context. Entity Role Detection is the task of assigning such roles to the entities. Usually real-world entities are of types: person, location and organization etc. Roles could be considered as domain-dependent subtypes of these types. In the cases, where retrieving a subset of entities based on their roles is needed, poses the problem of defining the role and entities having those roles. This paper presents the study of study of solving Entity Role Detection problem by modeling it as Named Entity Recognition (NER) and Entity Retrieval/Ranking task. In NER, these roles could be considered as mutually exclusive classes and standard NER methods like sequence tagging could be used. For Entity Retrieval, Roles could be formulated as Query and entities as Collection on which the query needs to be executed. The aspect of Entity Retrieval task, which is different than document retrieval task is that the entities and roles against which they need to be retrieved are indirectly described. We have formulated automated ways of learning representative words and phrases and building representations of roles and entities using them. We have also explored different contexts like sentence and document. Since the roles depend upon context, so it is not always possible to have large domain-specific dataset or knowledge bases for learning purposes, so we have tried to exploit the information from small dataset in domain-agnostic way.


Random Window Augmentations for Deep Learning Robustness in CT and Liver Tumor Segmentation

Østmo, Eirik A., Wickstrøm, Kristoffer K., Radiya, Keyur, Kampffmeyer, Michael C., Mikalsen, Karl Øyvind, Jenssen, Robert

arXiv.org Artificial Intelligence

Abstract--Contrast-enhanced Computed T omography (CT) is important for diagnosis and treatment planning for various medical conditions. Deep learning (DL) based segmentation models may enable automated medical image analysis for detecting and delineating tumors in CT images, thereby reducing clinicians' workload. Achieving generalization capabilities in limited data domains, such as radiology, requires modern DL models to be trained with image augmentation. However, naively applying augmentation methods developed for natural images to CT scans often disregards the nature of the CT modality, where the intensities measure Hounsfield Units (HU) and have important physical meaning. This paper challenges the use of such intensity augmentations for CT imaging and shows that they may lead to artifacts and poor generalization. T o mitigate this, we propose a CT -specific augmentation technique, called Random windowing, that exploits the available HU distribution of intensities in CT images. Random windowing encourages robustness to contrast-enhancement and significantly increases model performance on challenging images with poor contrast or timing. We perform ablations and analysis of our method on multiple datasets, and compare to, and outperform, state-of-the-art alternatives, while focusing on the challenge of liver tumor segmentation. Computed Tomography (CT) is a cornerstone in the diagnosis and treatment planning of various health conditions [1]. In liver applications, contrast-enhanced CT imaging enables precise imaging for detection and delineation of tumors, facilitating effective intervention strategies. With the rapid advancement of Deep Learning (DL), the utilization of computer vision (CV) models has become increasingly prevalent for automating tasks in radiology [2]-[5].


AI-Enabled Capabilities to Facilitate Next-Generation Rover Surface Operations

Luna, Cristina, Field, Robert, Kay, Steven

arXiv.org Artificial Intelligence

Contemporary Mars rovers such as Curiosity and Perseverance operate at average speeds on the order of 4.2 cm/s, with daily traverses typically below 100 m [1]. These constraints stem from conservative operational approaches necessitated by communication delays, irreplaceable hardware, and limited onboard processing capabilities. The traditional Sense-Model-Plan-Act (SMPA) paradigm requires frequent stops for terrain analysis, preventing continuous motion and severely limiting mission scope and scientific return. Missions requiring long-range access to diverse geological targets (sample-return campaigns) are particularly affected by these mobility constraints [2]. Recent advances in computer vision (CV) algorithms, compact ML models, and space-qualified computing platforms offer a practical path to maintaining safety while increasing autonomy and traverse speeds. In this work, we present a set of AI-enabled systems developed under ESA contracts RAPID, FASTNAV, ViBEKO and AIAXR, and CISRU. These systems were validated in Mars-and Lunar-analogue field trials and demonstrate substantial improvements in mobility and perception accuracy. The contributions presented in this work are: (1) a far-obstacle detection component which facilitates continuous motion at speeds in excess of 1.0 m/s; (2) a coordination framework enabling multi-robot human-robot workflows for resource extraction and handling; and (3) a suite of terrain classification models for operations.


Transit for All: Mapping Equitable Bike2Subway Connection using Region Representation Learning

Namgung, Min, Lee, JangHyeon, Ding, Fangyi, Chiang, Yao-Yi

arXiv.org Artificial Intelligence

Ensuring equitable public transit access remains challenging, particularly in densely populated cities like New York City (NYC), where low-income and minority communities often face limited transit accessibility. Bike-sharing systems (BSS) can bridge these equity gaps by providing affordable first- and last-mile connections. However, strategically expanding BSS into underserved neighborhoods is difficult due to uncertain bike-sharing demand at newly planned ("cold-start") station locations and limitations in traditional accessibility metrics that may overlook realistic bike usage potential. We introduce Transit for All (TFA), a spatial computing framework designed to guide the equitable expansion of BSS through three components: (1) spatially-informed bike-sharing demand prediction at cold-start stations using region representation learning that integrates multimodal geospatial data, (2) comprehensive transit accessibility assessment leveraging our novel weighted Public Transport Accessibility Level (wPTAL) by combining predicted bike-sharing demand with conventional transit accessibility metrics, and (3) strategic recommendations for new bike station placements that consider potential ridership and equity enhancement. Using NYC as a case study, we identify transit accessibility gaps that disproportionately impact low-income and minority communities in historically underserved neighborhoods. Our results show that strategically placing new stations guided by wPTAL notably reduces disparities in transit access related to economic and demographic factors. From our study, we demonstrate that TFA provides practical guidance for urban planners to promote equitable transit and enhance the quality of life in underserved urban communities.


IoT-based Noise Monitoring using Mobile Nodes for Smart Cities

Manthina, Bhima Sankar, Gujar, Shreyash, Chaudhari, Sachin, Vemuri1, Kavita, Chhirolya, Shivam

arXiv.org Artificial Intelligence

--Urban noise pollution poses a significant threat to public health, yet existing monitoring infrastructures offer limited spatial coverage and adaptability. This paper presents a scalable, low-cost, IoT -based, real-time environmental noise monitoring solution using mobile nodes ( sensor nodes on a moving vehicle). The system utilizes a low-cost sound sensor integrated with GPS-enabled modules to collect geotagged noise data at one-second intervals. The sound nodes are calibrated against a reference sound level meter in a laboratory setting to ensure accuracy using various machine learning (ML) algorithms such as Simple Linear Regression (SLR), Multiple Linear Regression (MLR), Polynomial Regression (PR), Segmented Regression (SR), Support V ector Regression (SVR), Decision Tree (DT), and Random Forest Regression (RFR). While laboratory calibration demonstrates high accuracy, it is shown that the performance of the nodes degrades during data collection in a moving vehicle. T o address this, it is demonstrated that the calibration must be performed on the IoT -based node based on the data collected in a moving environment along with the reference device. The system was deployed in Hyderabad, India, through three measurement campaigns across 27 days, capturing 436,420 data points. Results highlight temporal and spatial noise variations across weekdays, weekends, and during Diwali. Incorporating vehicular velocity into the calibration significantly improves accuracy. The proposed system demonstrates the potential for widespread deployment of IoT -based noise sensing networks in smart cities, enabling effective noise pollution management and urban planning. OISE pollution, also known as environmental noise or sound pollution, refers to unwanted or excessive sound that disrupts human activities and negatively impacts human health [1]. The known sources of noise pollution include transportation (such as road traffic), industrial activities, construction, and urban crowding [2].


ACCESS-AV: Adaptive Communication-Computation Codesign for Sustainable Autonomous Vehicle Localization in Smart Factories

Bhattacharjya, Rajat, Sarkar, Arnab, Kool, Ish, Baidya, Sabur, Dutt, Nikil

arXiv.org Artificial Intelligence

Autonomous Delivery Vehicles (ADVs) are increasingly used for transporting goods in 5G network-enabled smart factories, with the compute-intensive localization module presenting a significant opportunity for optimization. We propose ACCESS-AV, an energy-efficient Vehicle-to-Infrastructure (V2I) localization framework that leverages existing 5G infrastructure in smart factory environments. By opportunistically accessing the periodically broadcast 5G Synchronization Signal Blocks (SSBs) for localization, ACCESS-AV obviates the need for dedicated Roadside Units (RSUs) or additional onboard sensors to achieve energy efficiency as well as cost reduction. We implement an Angle-of-Arrival (AoA)-based estimation method using the Multiple Signal Classification (MUSIC) algorithm, optimized for resource-constrained ADV platforms through an adaptive communication-computation strategy that dynamically balances energy consumption with localization accuracy based on environmental conditions such as Signal-to-Noise Ratio (SNR) and vehicle velocity. Experimental results demonstrate that ACCESS-AV achieves an average energy reduction of 43.09% compared to non-adaptive systems employing AoA algorithms such as vanilla MUSIC, ESPRIT, and Root-MUSIC. It maintains sub-30 cm localization accuracy while also delivering substantial reductions in infrastructure and operational costs, establishing its viability for sustainable smart factory environments.


Developing Lightweight DNN Models With Limited Data For Real-Time Sign Language Recognition

Nikitin, Nikita, Fomin, Eugene

arXiv.org Artificial Intelligence

We present a novel framework for real-time sign language recognition using lightweight DNNs trained on limited data. Our system addresses key challenges in sign language recognition, including data scarcity, high computational costs, and discrepancies in frame rates between training and inference environments. By encoding sign language specific parameters, such as handshape, palm orientation, movement, and location into vectorized inputs, and leveraging MediaPipe for landmark extraction, we achieve highly separable input data representations. Our DNN architecture, optimized for sub 10MB deployment, enables accurate classification of 343 signs with less than 10ms latency on edge devices. The data annotation platform 'slait data' facilitates structured labeling and vector extraction. Our model achieved 92% accuracy in isolated sign recognition and has been integrated into the 'slait ai' web application, where it demonstrates stable inference.


Don't lie to your friends: Learning what you know from collaborative self-play

Eisenstein, Jacob, Aghajani, Reza, Fisch, Adam, Dua, Dheeru, Huot, Fantine, Lapata, Mirella, Zayats, Vicky, Berant, Jonathan

arXiv.org Artificial Intelligence

To be helpful assistants, AI agents must be aware of their own capabilities and limitations. This includes knowing when to answer from parametric knowledge versus using tools, when to trust tool outputs, and when to abstain or hedge. Such capabilities are hard to teach through supervised fine-tuning because they require constructing examples that reflect the agent's specific capabilities. We therefore propose a radically new approach to teaching agents what they know: \emph{collaborative self-play}. We construct multi-agent collaborations in which the group is rewarded for collectively arriving at correct answers. The desired meta-knowledge emerges from the incentives built into the structure of the interaction. We focus on small societies of agents that have access to heterogeneous tools (corpus-specific retrieval), and therefore must collaborate to maximize their success while minimizing their effort. Experiments show that group-level rewards for multi-agent communities can induce policies that \emph{transfer} to improve tool use and selective prediction in settings where individual agents are deployed in isolation.